81 research outputs found
Parallel Exhaustive Search without Coordination
We analyze parallel algorithms in the context of exhaustive search over
totally ordered sets. Imagine an infinite list of "boxes", with a "treasure"
hidden in one of them, where the boxes' order reflects the importance of
finding the treasure in a given box. At each time step, a search protocol
executed by a searcher has the ability to peek into one box, and see whether
the treasure is present or not. By equally dividing the workload between them,
searchers can find the treasure times faster than one searcher.
However, this straightforward strategy is very sensitive to failures (e.g.,
crashes of processors), and overcoming this issue seems to require a large
amount of communication. We therefore address the question of designing
parallel search algorithms maximizing their speed-up and maintaining high
levels of robustness, while minimizing the amount of resources for
coordination. Based on the observation that algorithms that avoid communication
are inherently robust, we analyze the best running time performance of
non-coordinating algorithms. Specifically, we devise non-coordinating
algorithms that achieve a speed-up of for two searchers, a speed-up of
for three searchers, and in general, a speed-up of
for any searchers. Thus, asymptotically, the speed-up is only four
times worse compared to the case of full-coordination, and our algorithms are
surprisingly simple and hence applicable. Moreover, these bounds are tight in a
strong sense as no non-coordinating search algorithm can achieve better
speed-ups. Overall, we highlight that, in faulty contexts in which coordination
between the searchers is technically difficult to implement, intrusive with
respect to privacy, and/or costly in term of resources, it might well be worth
giving up on coordination, and simply run our non-coordinating exhaustive
search algorithms
High Performance Computing Applications in Remote Sensing Studies for Land Cover Dynamics
Global and regional land cover studies require the ability to apply complex models on selected subsets of large amounts of multi-sensor and multi-temporal data sets that have been derived from raw instrument measurements using widely accepted pre-processing algorithms. The computational and storage requirements of most such studies far exceed what is possible on a single workstation environment. We have been pursuing a new approach that couples scalable and open distributed heterogeneous hardware with the development of high performance software for processing, indexing, and organizing remotely sensed data. Hierarchical data management tools are used to ingest raw data, create metadata, and organize the archived data so as to automatically achieve computational load balancing among the available nodes and minimize I/O overheads. We illustrate our approach with four specific examples. The first is the development of the first fast operational scheme for the atmospheric correction of Landsat TM scenes, while the second example focuses on image segmentation using a novel hierarchical connected components algorithm. Retrieval of global BRDF (Bidirectional Reflectance Distribution Function) in the red and near infrared wavelengths using four years (1983 to 1986) of Pathfinder AVHRR Land (PAL) data set is the focus of our third example. The fourth example is the development of a hierarchical data organization scheme that allows on-demand processing and retrieval of regional and global AVHRR data sets. Our results show that substantial improvements in computational times can be achieved by using the high performance computing technology
The accuracy of breast volume measurement methods: a systematic review
Breast volume is a key metric in breast surgery and there are a number of different methods which measure it. However, a lack of knowledge regarding a method’s accuracy and comparability has made it difficult to establish a clinical standard. We have performed a systematic review of the literature to examine the various techniques for measurement of breast volume and to assess their accuracy and usefulness in clinical practice. Each of the fifteen studies we identified had more than ten live participants and assessed volume measurement accuracy using a gold-standard based on the volume, or mass, of a mastectomy specimen. Many of the studies from this review report large (> 200 ml) uncertainty in breast volume and many fail to assess measurement accuracy using appropriate statistical tools. Of the methods assessed, MRI scanning consistently demonstrated the highest accuracy with three studies reporting errors lower than 10% for small (250 ml), medium (500 ml) and large (1,000 ml) breasts. However, as a high-cost, non-routine assessment other methods may be more appropriate
Balancing Minimum Spanning and Shortest Path Trees
This paper give a simple linear-time algorithm that, given a weighted
digraph, finds a spanning tree that simultaneously approximates a shortest-path
tree and a minimum spanning tree. The algorithm provides a continuous
trade-off: given the two trees and epsilon > 0, the algorithm returns a
spanning tree in which the distance between any vertex and the root of the
shortest-path tree is at most 1+epsilon times the shortest-path distance, and
yet the total weight of the tree is at most 1+2/epsilon times the weight of a
minimum spanning tree. This is the best tradeoff possible. The paper also
describes a fast parallel implementation.Comment: conference version: ACM-SIAM Symposium on Discrete Algorithms (1993
Computing discriminating and generic words
International audienceWe study the following three problems of computing generic or discriminating words for a given collection of documents. Given a pattern P and a threshold d, we want to report (i) all longest extensions of P which occur in at least d documents, (ii) all shortest extensions of P which occur in less than d documents, and (iii) all shortest extensions of P which occur only in d selected documents. For these problems, we propose efficient algorithms based on suffix trees and using advanced data structure techniques. For problem (i), we propose an optimal solution with constant running time per output word
Succinct Indices for Range Queries with applications to Orthogonal Range Maxima
We consider the problem of preprocessing points in 2D, each endowed with
a priority, to answer the following queries: given a axis-parallel rectangle,
determine the point with the largest priority in the rectangle. Using the ideas
of the \emph{effective entropy} of range maxima queries and \emph{succinct
indices} for range maxima queries, we obtain a structure that uses O(N) words
and answers the above query in time. This is a direct
improvement of Chazelle's result from FOCS 1985 for this problem -- Chazelle
required words to answer queries in
time for any constant .Comment: To appear in ICALP 201
- …